Picture for Tushar Krishna

Tushar Krishna

NSFlow: An End-to-End FPGA Framework with Scalable Dataflow Architecture for Neuro-Symbolic AI

Add code
Apr 29, 2025
Viaarxiv icon

Towards Easy and Realistic Network Infrastructure Testing for Large-scale Machine Learning

Add code
Apr 29, 2025
Viaarxiv icon

Generative AI in Embodied Systems: System-Level Analysis of Performance, Efficiency and Scalability

Add code
Apr 26, 2025
Viaarxiv icon

Accelerating LLM Inference with Flexible N:M Sparsity via A Fully Digital Compute-in-Memory Accelerator

Add code
Apr 19, 2025
Viaarxiv icon

Understanding and Optimizing Multi-Stage AI Inference Pipelines

Add code
Apr 16, 2025
Viaarxiv icon

OuroMamba: A Data-Free Quantization Framework for Vision Mamba Models

Add code
Mar 13, 2025
Viaarxiv icon

AIRCHITECT v2: Learning the Hardware Accelerator Design Space through Unified Representations

Add code
Jan 17, 2025
Viaarxiv icon

Leveraging ASIC AI Chips for Homomorphic Encryption

Add code
Jan 13, 2025
Viaarxiv icon

TURBOATTENTION: Efficient Attention Approximation For High Throughputs LLMs

Add code
Dec 11, 2024
Figure 1 for TURBOATTENTION: Efficient Attention Approximation For High Throughputs LLMs
Figure 2 for TURBOATTENTION: Efficient Attention Approximation For High Throughputs LLMs
Figure 3 for TURBOATTENTION: Efficient Attention Approximation For High Throughputs LLMs
Figure 4 for TURBOATTENTION: Efficient Attention Approximation For High Throughputs LLMs
Viaarxiv icon

MicroScopiQ: Accelerating Foundational Models through Outlier-Aware Microscaling Quantization

Add code
Nov 08, 2024
Figure 1 for MicroScopiQ: Accelerating Foundational Models through Outlier-Aware Microscaling Quantization
Figure 2 for MicroScopiQ: Accelerating Foundational Models through Outlier-Aware Microscaling Quantization
Figure 3 for MicroScopiQ: Accelerating Foundational Models through Outlier-Aware Microscaling Quantization
Figure 4 for MicroScopiQ: Accelerating Foundational Models through Outlier-Aware Microscaling Quantization
Viaarxiv icon